人类可以通过最小的相互干扰连续学习几项任务,但一次接受多个任务进行培训时的表现较差。标准深神经网络相反。在这里,我们提出了针对人工神经网络的新型计算限制,灵感来自灵长类动物前额叶皮层的较​​早作品,以捕获交织训练的成本,并允许网络在不忘记的情况下按顺序学习两个任务。我们通过两个算法主题,所谓的“呆滞”任务单元和HEBBIAN训练步骤增强了标准随机梯度下降,该步骤加强了任务单元和编码与任务相关信息的隐藏单元之间的连接。我们发现,“缓慢”的单元在培训期间引入了转换成本,该单元在交错训练下偏向表示的表示,而忽略了上下文提示的联合表示,而Hebbian步骤则促进了从任务单元到隐藏层的门控方案的形成这会产生正交表示,完全防止干扰。在先前发布的人类行为数据上验证该模型表明,它与接受过封锁或交错课程训练的参与者的表现相匹配,并且这些绩效差异是由真实类别边界的误解驱动的。
translated by 谷歌翻译
Late-life depression (LLD) is a highly prevalent mood disorder occurring in older adults and is frequently accompanied by cognitive impairment (CI). Studies have shown that LLD may increase the risk of Alzheimer's disease (AD). However, the heterogeneity of presentation of geriatric depression suggests that multiple biological mechanisms may underlie it. Current biological research on LLD progression incorporates machine learning that combines neuroimaging data with clinical observations. There are few studies on incident cognitive diagnostic outcomes in LLD based on structural MRI (sMRI). In this paper, we describe the development of a hybrid representation learning (HRL) framework for predicting cognitive diagnosis over 5 years based on T1-weighted sMRI data. Specifically, we first extract prediction-oriented MRI features via a deep neural network, and then integrate them with handcrafted MRI features via a Transformer encoder for cognitive diagnosis prediction. Two tasks are investigated in this work, including (1) identifying cognitively normal subjects with LLD and never-depressed older healthy subjects, and (2) identifying LLD subjects who developed CI (or even AD) and those who stayed cognitively normal over five years. To the best of our knowledge, this is among the first attempts to study the complex heterogeneous progression of LLD based on task-oriented and handcrafted MRI features. We validate the proposed HRL on 294 subjects with T1-weighted MRIs from two clinically harmonized studies. Experimental results suggest that the HRL outperforms several classical machine learning and state-of-the-art deep learning methods in LLD identification and prediction tasks.
translated by 谷歌翻译
Since early in the coronavirus disease 2019 (COVID-19) pandemic, there has been interest in using artificial intelligence methods to predict COVID-19 infection status based on vocal audio signals, for example cough recordings. However, existing studies have limitations in terms of data collection and of the assessment of the performances of the proposed predictive models. This paper rigorously assesses state-of-the-art machine learning techniques used to predict COVID-19 infection status based on vocal audio signals, using a dataset collected by the UK Health Security Agency. This dataset includes acoustic recordings and extensive study participant meta-data. We provide guidelines on testing the performance of methods to classify COVID-19 infection status based on acoustic features and we discuss how these can be extended more generally to the development and assessment of predictive methods based on public health datasets.
translated by 谷歌翻译
To analyze this characteristic of vulnerability, we developed an automated deep learning method for detecting microvessels in intravascular optical coherence tomography (IVOCT) images. A total of 8,403 IVOCT image frames from 85 lesions and 37 normal segments were analyzed. Manual annotation was done using a dedicated software (OCTOPUS) previously developed by our group. Data augmentation in the polar (r,{\theta}) domain was applied to raw IVOCT images to ensure that microvessels appear at all possible angles. Pre-processing methods included guidewire/shadow detection, lumen segmentation, pixel shifting, and noise reduction. DeepLab v3+ was used to segment microvessel candidates. A bounding box on each candidate was classified as either microvessel or non-microvessel using a shallow convolutional neural network. For better classification, we used data augmentation (i.e., angle rotation) on bounding boxes with a microvessel during network training. Data augmentation and pre-processing steps improved microvessel segmentation performance significantly, yielding a method with Dice of 0.71+/-0.10 and pixel-wise sensitivity/specificity of 87.7+/-6.6%/99.8+/-0.1%. The network for classifying microvessels from candidates performed exceptionally well, with sensitivity of 99.5+/-0.3%, specificity of 98.8+/-1.0%, and accuracy of 99.1+/-0.5%. The classification step eliminated the majority of residual false positives, and the Dice coefficient increased from 0.71 to 0.73. In addition, our method produced 698 image frames with microvessels present, compared to 730 from manual analysis, representing a 4.4% difference. When compared to the manual method, the automated method improved microvessel continuity, implying improved segmentation performance. The method will be useful for research purposes as well as potential future treatment planning.
translated by 谷歌翻译
在许多情况下,更简单的模型比更复杂的模型更可取,并且该模型复杂性的控制是机器学习中许多方法的目标,例如正则化,高参数调整和体系结构设计。在深度学习中,很难理解复杂性控制的潜在机制,因为许多传统措施并不适合深度神经网络。在这里,我们开发了几何复杂性的概念,该概念是使用离散的dirichlet能量计算的模型函数变异性的量度。使用理论论据和经验结果的结合,我们表明,许多常见的训练启发式方法,例如参数规范正规化,光谱规范正则化,平稳性正则化,隐式梯度正则化,噪声正则化和参数初始化的选择,都可以控制几何学复杂性,并提供一个统一的框架,以表征深度学习模型的行为。
translated by 谷歌翻译
步态冻结(FOG)是帕金森氏病的最常见症状之一,这是中枢神经系统的神经退行性疾病,影响了世界各地数百万的人。为了满足提高雾的治疗质量的紧迫需求,设计雾计算机辅助检测和量化工具的需求越来越重要。作为一种用于收集运动模式的非侵入性技术,从压力敏感步态垫中获得的脚步压力序列为评估诊所和家庭环境中的雾气提供了绝佳的机会。在这项研究中,提出了雾检测为一项顺序建模任务,并提出了一种新颖的深度学习结构,即对对抗性时空网络(ASTN),提出了跨多个级别的雾模式。引入了一种新型的对抗训练方案,并具有多级主题鉴别器,以获得独立的雾代表示,这有助于降低由于高主体间方差而导致的过度拟合风险。结果,对于看不见的受试者,可以实现强大的雾检测。拟议的计划还阐明了从其他场景中改善主题级临床研究,因为它可以与许多现有的深层建筑集成在一起。据我们所知,这是基于脚步压力的雾检测的最早研究之一,利用ASTN的方法是追求独立于主题的表示形式的第一个深神经网络架构。从21名受试者收集的393次试验的实验结果表明,AUC 0.85的雾检测提出的ASTN表现令人鼓舞。
translated by 谷歌翻译
在本文中,使用聚类和阈值算法实现了DIBA数据集细菌属和物种的半自动注释。深度学习模型经过训练,以实现细菌物种的语义分割和分类。分类精度达到95%。深度学习模型在生物医学图像处理中发现了巨大的应用。从革兰氏阴性微观图像中自动分割细菌对于诊断呼吸道和尿路感染,检测癌症等至关重要。深度学习将有助于生物学家在更少的时间内获得可靠的结果。此外,可以减少许多人类干预措施。这项工作可能有助于检测尿液涂片图像,痰液涂片图像等的细菌,以诊断尿路感染,结核病,肺炎等。
translated by 谷歌翻译
唇裂是一种先天性异常,需要专家手术修复。外科医生必须具有丰富的经验和理论知识才能进行手术,并且已经提出了人工智能(AI)方法来指导外科医生改善手术结局。如果可以使用AI来预测修复的唇唇的外观,那么外科医生可以将其用作辅助手术技术来调整其手术技术并改善结果。为了在保护患者隐私时探索这个想法的可行性,我们提出了一种基于深度学习的图像镶嵌方法,该方法能够覆盖唇裂,并产生唇彩,而无需裂缝。我们的实验是在两个现实世界中的裂口数据集上进行的,并由专家cleft唇外科医生评估,以证明该方法的可行性。
translated by 谷歌翻译
我们解决以下动作效应预测任务。给定描绘世界初始状态和文本中表达的动作的图像,预测了动作后描绘世界状态的图像。预测应具有与输入图像相同的场景上下文。我们探讨了最近提出的GLIDE模型执行此任务的使用。Glide是一个生成性神经网络,可以合成图像的掩盖区域(涂层),以短片段为条件。我们的想法是掩盖预期动作效果的输入图像的区域。然后使用滑行以在所需动作为条件的蒙面区域内涂抹涂漆。这样,结果图像具有与输入图像相同的背景上下文,并更新以显示动作的效果。我们使用带有动作标记的自我中心视频的Epic数据集给出了实验的定性结果。
translated by 谷歌翻译
可穿戴机器人设备有可能协助和保护用户。为了设计智能头盔,本文研究了音频和视觉警告的有效性,以帮助参与者振作起来。一项用户研究检查了运行时对用户应用的不同警告和影响。从不同的方向应用了缩放到用户质量的扰动力,并测量用户位移以表征警告的有效性。这是使用适应于运动循环期间精确矩,向前,向后,右或左侧扰动力来向前,向后,右或左侧扰动力进行的踏板活动的活动风洞来完成的。本文介绍了该系统的概述,并展示了步态过程中精确发出一致警告和扰动的能力。用户研究结果突出了视觉和音频警告的有效性,以帮助用户振作起来,从而导致指南,从而为未来的人类机器人警告系统提供信息。
translated by 谷歌翻译